# Multilingual mixed training
Gemma 2 Llama Swallow 9b It V0.1
The Gemma-2-Llama-Swallow series of models are multilingual large models constructed through continuous pre-training based on Gemma-2, with a particular enhancement in Japanese ability.
Large Language Model
Transformers Supports Multiple Languages

G
tokyotech-llm
2,491
3
Stockmark 2 100B Instruct Beta
MIT
Stockmark-2-100B is a 100-billion parameter large language model focused on Japanese capabilities, pre-trained on 1.5 trillion tokens of multilingual data and enhanced with Japanese synthetic data for improved instruction following.
Large Language Model
Transformers Supports Multiple Languages

S
stockmark
1,004
9
Swiss German Xlm Roberta Base
This model is a multilingual model fine-tuned on Swiss German text data through continual pre-training based on xlm-roberta-base, supporting both Swiss German and Standard German processing.
Large Language Model
Transformers Other

S
ZurichNLP
17
2
Openhathi 7B Hi V0.1 Base
The first model in the OpenHathi series, based on the Llama2 architecture, supports Hindi, English, and mixed languages, with 7 billion parameters.
Large Language Model Other
O
sarvamai
655
112
Byt5 Korean Base
Apache-2.0
ByT5-Korean is a customized Korean extension of Google's ByT5, specifically optimized for Korean syllable encoding processing.
Large Language Model
Transformers

B
everdoubling
55
2
Featured Recommended AI Models